National Repository of Grey Literature 26 records found  1 - 10nextend  jump to record: Search took 0.01 seconds. 
Depth of variance matrices
Brabenec, Tomáš ; Nagy, Stanislav (advisor) ; Hlubinka, Daniel (referee)
The scatter halfspace depth is a quite recently established concept which extends the idea of the location halfspace depth for positive definite matrices. It provides an interest- ing insight into the problem of suitability quantification of a matrix for the description of the covariance structure of the multivariate distribution. The thesis focuses on the investigation of theoretical properties of the depth for both general and more specific probability distributions which can be used for data analysis. It turns out that the es- timators of scatter parameters based on the empirical scatter depth are quite effective even under relatively weak assumptions. These estimators are useful especially for dealing with a sample containing outliers or contaminating observations. 1
Weighted Halfspace Depths and Their Properties
Kotík, Lukáš
Statistical depth functions became well known nonparametric tool of multivariate data analyses. The most known depth functions include the halfspace depth. Although the halfspace depth has many desirable properties, some of its properties may lead to biased and misleading results especially when data are not elliptically symmetric. The thesis introduces 2 new classes of the depth functions. Both classes generalize the halfspace depth. They keep some of its properties and since they more respect the geometric structure of data they usually lead to better results when we deal with non-elliptically symmetric, multimodal or mixed distributions. The idea presented in the thesis is based on replacing the indicator of a halfspace by more general weight function. This provides us with a continuum, especially if conic-section weight functions are used, between a local view of data (e.g. kernel density estimate) and a global view of data as is e.g. provided by the halfspace depth. The rate of localization is determined by the choice of the weight functions and theirs parameters. Properties including the uniform strong consistency of the proposed depth functions are proved in the thesis. Limit distribution is also discussed together with some other data depth related topics (regression depth, functional data depth)...
Nonparametric Bootstrap Techniques for Implicitly Weighted Robust Estimators
Kalina, Jan
The paper is devoted to highly robust statistical estimators based on implicit weighting, which have a potential to find econometric applications. Two particular methods include a robust correlation coefficient based on the least weighted squares regression and the minimum weighted covariance determinant estimator, where the latter allows to estimate the mean and covariance matrix of multivariate data. New tools are proposed allowing to test hypotheses about these robust estimators or to estimate their variance. The techniques considered in the paper include resampling approaches with or without replacement, i.e. permutation tests, bootstrap variance estimation, and bootstrap confidence intervals. The performance of the newly described tools is illustrated on numerical examples. They reveal the suitability of the robust procedures also for non-contaminated data, as their confidence intervals are not much wider compared to those for standard maximum likelihood estimators. While resampling without replacement turns out to be more suitable for hypothesis testing, bootstrapping with replacement yields reliable confidence intervals but not corresponding hypothesis tests.
Nonparametric Bootstrap Techniques for Implicitly Weighted Robust Estimators
Kalina, Jan
The paper is devoted to highly robust statistical estimators based on implicit weighting, which have a potential to find econometric applications. Two particular methods include a robust correlation coefficient based on the least weighted squares regression and the minimum weighted covariance determinant estimator, where the latter allows to estimate the mean and covariance matrix of multivariate data. New tools are proposed allowing to test hypotheses about these robust estimators or to estimate their variance. The techniques considered in the paper include resampling approaches with or without replacement, i.e. permutation tests, bootstrap variance estimation, and bootstrap confidence intervals. The performance of the newly described tools is illustrated on numerical examples. They reveal the suitability of the robust procedures also for non-contaminated data, as their confidence intervals are not much wider compared to those for standard maximum likelihood estimators. While resampling without replacement turns out to be more suitable for hypothesis testing, bootstrapping with replacement yields reliable confidence intervals but not corresponding hypothesis tests.
Weighted Halfspace Depths and Their Properties
Kotík, Lukáš
Statistical depth functions became well known nonparametric tool of multivariate data analyses. The most known depth functions include the halfspace depth. Although the halfspace depth has many desirable properties, some of its properties may lead to biased and misleading results especially when data are not elliptically symmetric. The thesis introduces 2 new classes of the depth functions. Both classes generalize the halfspace depth. They keep some of its properties and since they more respect the geometric structure of data they usually lead to better results when we deal with non-elliptically symmetric, multimodal or mixed distributions. The idea presented in the thesis is based on replacing the indicator of a halfspace by more general weight function. This provides us with a continuum, especially if conic-section weight functions are used, between a local view of data (e.g. kernel density estimate) and a global view of data as is e.g. provided by the halfspace depth. The rate of localization is determined by the choice of the weight functions and theirs parameters. Properties including the uniform strong consistency of the proposed depth functions are proved in the thesis. Limit distribution is also discussed together with some other data depth related topics (regression depth, functional data depth)...
Weighted Halfspace Depths and Their Properties
Kotík, Lukáš ; Hlubinka, Daniel (advisor) ; Omelka, Marek (referee) ; Mosler, Karl (referee)
Statistical depth functions became well known nonparametric tool of multivariate data analyses. The most known depth functions include the halfspace depth. Although the halfspace depth has many desirable properties, some of its properties may lead to biased and misleading results especially when data are not elliptically symmetric. The thesis introduces 2 new classes of the depth functions. Both classes generalize the halfspace depth. They keep some of its properties and since they more respect the geometric structure of data they usually lead to better results when we deal with non-elliptically symmetric, multimodal or mixed distributions. The idea presented in the thesis is based on replacing the indicator of a halfspace by more general weight function. This provides us with a continuum, especially if conic-section weight functions are used, between a local view of data (e.g. kernel density estimate) and a global view of data as is e.g. provided by the halfspace depth. The rate of localization is determined by the choice of the weight functions and theirs parameters. Properties including the uniform strong consistency of the proposed depth functions are proved in the thesis. Limit distribution is also discussed together with some other data depth related topics (regression depth, functional data depth)...
Some Robust Distances for Multivariate Data
Kalina, Jan ; Peštová, Barbora
Numerous methods of multivariate statistics and data mining suffer from the presence of outlying measurements in the data. This paper presents new distance measures suitable for continuous data. First, we consider a Mahalanobis distance suitable for high-dimensional data with the number of variables (largely) exceeding the number of observations. We propose its doubly regularized version, which combines a regularization of the covariance matrix with replacing the means of multivariate data by their regularized counterparts. We formulate explicit expressions for some versions of the regularization of the means, which can be interpreted as a denoising (i.e. robust version) of standard means. Further, we propose a robust cosine similarity measure, which is based on implicit weighting of individual observations. We derive properties of the newly proposed robust cosine similarity, which includes a proof of the high robustness in terms of the breakdown point.
Volatility of selected separators/classifiers wrt. data sets from field of particle physics
Jiřina, Marcel ; Hakl, František
We study the volatility, i.e. influence of random changes in data sets to overall separation/classification behavior of separators/classifiers. This is motivated by the fact, that simulated data and true data from ATLAS experiment may differ, and a question arises what if separators or cuts are optimized for simulated data, and then used for true data from the experiment. This behavior was studied using simulated data modified by artificial distortions of known size. We found that even slight change in data sets causes a little worse result than supposed but, surprisingly, even relatively large distortions give then nearly the same results. Only truly great variations cause degradation of separation quality of separator/classifier as well as of the cuts method.
Fulltext: content.csg - Download fulltextPDF
Plný tet: v1126-11 - Download fulltextPDF

National Repository of Grey Literature : 26 records found   1 - 10nextend  jump to record:
Interested in being notified about new results for this query?
Subscribe to the RSS feed.